- Thursday 31st January: linux.conf.au keynote,
- Monday 4th February: Public lecture at the University of Melbourne,
- Tuesday 5th February: Public lecture, Sydney Town Hall.
Saturday, December 15, 2012
Tim Berners-Lee at linux.conf.au in Canberra 31 January
Wednesday, May 25, 2011
Metadata, privacy and information policy
Professor McMillan also pointed out that metadata is important to the mechanics of implementing the government's information policy. He held up a copy of the new "Principles on open public sector information", which were released today. (along with a "Report on review and development of principles". It is good to see that the commission released the documents in the form of simple and easy to read HTML files, as well as PDF and RTF. They also put the HTML version first, which will be most useful.
One of the audience asked about intellectual property. The Commissioner replied this was the responsibility of the Attorney General's department, but pointed out this was touched on in the information principles and AGs recommend use of a Creative Commons licence for material to be released to the public.
Here are the eight principles of open government sector information:
Principle 1: Open access to information - a default position
Information held by Australian Government agencies is a valuable national resource. If there is no legal need to protect the information it should be open to public access. Information publication enhances public access. Agencies should use information technology to disseminate public sector information, applying a presumption of openness and adopting a proactive publication stance.
Principle 2: Engaging the community
Australian Government policy requires agencies to engage the community online in policy design and service delivery. This should apply to agency information publication practices. Agencies should:
- consult the community in deciding what information to publish and about agency publication practices
- welcome community feedback about the quality, completeness, usefulness and accuracy of published information
- respond promptly to comments received from the community and to requests for information
- employ Web 2.0 tools to support community consultation.
Principle 3: Effective information governance
Australian Government agencies should manage information as a core strategic asset. A senior executive ‘information champion' or knowledge officer in the agency should be responsible for information management and governance, including:
- providing leadership on agency compliance with the Information Publication Scheme and Disclosure Log
- ensuring agency compliance with legislative and policy requirements on information management and publication
- managing agency information to ensure its integrity, security and accessibility
- instigating strategic planning on information resource management
- ensuring community consultation on agency information policy and publication practices.
The senior officer should be supported by an information governance body that may include people from outside the agency.
Principle 4: Robust information asset management
Effective information management requires agencies to:
- maintain an asset inventory or register of the agency's information
- identify the custodian of each information holding and the responsibilities of that officer
- train staff in information management
- establish clear procedures and lines of authority for decisions on information publication and release
- decide if information should be prepared for publication at the time it is created and the form of publication
- document known limitations on data quality
- identify data that must be managed in accordance with legislative and legal requirements, including requirements relating to data security and protection of personal information, intellectual property, business confidentiality and legal professional privilege
- protect information against inappropriate or unauthorised use, access or disclosure
- preserve information for an appropriate period of time based on sound archival practices.
Principle 5: Discoverable and useable information
The economic and social value of public sector information can be enhanced by publication and information sharing. This requires that information can easily be discovered and used by the community and other stakeholders. To support this objective agencies should:
- publish an up to date information asset register
- ensure that information published online is in an open and standards-based format and is machine-readable
- attach high quality metadata to information so that it can be easily located and linked to similar information using standard web search applications
- publish information in accordance with the Web Content Accessibility Guidelines version 2 (WCAG 2.0) endorsed by the Australian Government in November 2009.
Principle 6: Clear reuse rights
The economic and social value of public sector information is enhanced when it is made available for reuse on open licensing terms. The Guidelines on Licensing Public Sector Information for Australian Government Agencies require agencies to decide licensing conditions when publishing information online. The default condition should be the Creative Commons BY standard, as recommended in the Intellectual Property Principles for Australian Government Agencies, that apply to agencies subject to the Financial and Management Accountability Act 1997. Additional guidance on selecting an appropriate licence is given in the Australian Government Open Access and Licensing Framework (AUSGOAL).
Principle 7: Appropriate charging for access
The FOI Act requires agencies to facilitate public access to information at the lowest reasonable cost. This principle applies when information is provided upon request or is published by an agency. Other Acts also authorise charges for specific documents or information access.
Agencies can reduce the cost of public access by publishing information online, especially information that is routinely sought by the public. Charges that may be imposed by an agency for providing access should be clearly explained in an agency policy that is published and regularly reviewed.
Principle 8: Transparent enquiry and complaints processes
Agency decision making about information publication should be transparent. This can be supported, within the agency's information governance framework, by an enquiry and complaints procedure for the public to raise issues about agency publication and access decisions. The procedure should be published, explain how enquiries and complaints will be handled, set timeframes for responding, identify possible remedies and complaint outcomes, and require that written reasons be provided in complaint resolution. ...
From: "Principles on open public sector information", AOIC, 25 May 2011
Publishing BBC Metadata on the Web
Tom demonstrated the BBC Nature website, which in addition to ordinary web pages, provides structured data, using RSS and RDF and semantic mark-up using microformats. This data is available for others to use and is also used by the BBC to create new stories.
Tom also mentioned dbpedia, an attempt to structure Wikipedia data. At this point he argued that there is no metadata and what is commonly though of is data is actually metadata. In a reference to Stephen Hawking, Tom said "Turtles all the way down". This is an metaphor for infinte recursion, however, I would argue it is "metadata all the way down". James Gleick argues in his book "The Information: A History, a Theory, a Flood", that the ability to reason abstractly came after writing ("if all horses are white ..."). That seems unlikely, as I am sure horse breeders reasoned on the nature of a good horse, before written language. Data and metadata are intertwined by their nature, not due to a human invention.
Tom argued that we needed to move from the document web to the data web, the web of things, which is what the semantic web is for. However, after spending many years trying to understand the semantic web and teach it to university students (supervising several masters students doing project on using it for cataloguing indigenous cultural material), I think this is a concept which needs to be further refined and simplified to be widely used. Tim Berners-Lee's key contribution with the World Wide Web was to take an existing complex electronic document standard (SGML) and simplify it to make something easy enough to use (HTML). Ever since, information professionals have argued that HTML is flawed, some tinkered with SGML and produced XML, others tinkered with HTML to make XHTML, but lost was the simplicity of HTML In my view the semantic web similarly needs simplification, even if the purists then say it is incomplete.
Tom then explained that the BBC use metadata for program guides. The importance is not the metadata but the information it describes. This is the key point which information professionals tend to find so obvious, that they forget to explain. While they may say metadata is data about data, but do not say why this is useful. That is a topic I will explore in my talk to the conference tomorrow, with Senator Lundy on "Designing for Democratic Dialogue: More than Mating iPads" (11.00 am on Thursday 26th May, 2011).
Next on the program today we have Greg Stone, Chief Technology Officer, Microsoft Australia and Professor John McMillan, Australian Information Commissioner, who is launching the new government information policy.
Wednesday, May 18, 2011
BBC on Semantic Web in Canberra
... The BBC is an acknowledged leader in semantic web publishing and the use of metadata to create dynamically linked sites. Tom Scott has executive responsibility for three of the broadcast giant's sites, which are often held up as exemplary projects within the Linked Data community. These are: the BBC Nature site (bbc.co.uk/nature) incorporating a major Linked Data and video publishing element; the BBC Programme Support site (bbc.co.uk/programmes), a website that publishes a page (URI and metadata) for every programme the BBC broadcasts; and the BBC's Music site (bbc.co.uk/music), a website that integrates into the BBC broadcast systems and BBC Programmes.
As an active blogger and expert in semantic web publishing, Tom will paint the picture of both the current and future demands of digital publishing and information management and how metadata can enable that using real live examples. “The idea is to help people think about the Web from a different perspective, one that is ultimately empowering because in doing so you can more easily deliver real benefits to your organisation and users” said Scott....
From: BBC Tom Scott to present at Meta2011, Press Release, IMM, 12 May 2011
Saturday, February 06, 2010
Wragge's identity finder
Wednesday, December 02, 2009
Connecting up online automatically
Some issues I can see with Dr. Fox's approach are those of security of the data and cost. Most e-science systems assume that all the data is available to anyone and is free. However, data access may need to be limited due to contractual agreements, privacy and national security. Also accessing the data and processing it may cost money. As a result one consideration in working out the answer to a question is what data you can get and what it will cost. Dr. Fox ended his talk by mentioning that educators should have access t summary data and that open source has potential, but it was not clear to me how this fitted with the discussion which went before.
As well as the implementation technology using ontologies, what I found of interest was a development methodology. This might have applicability to projects like the NBN, where there is a need to rapidly develop a system which negotiates between components owned by different organisations. It may also be useful for quasi-commercial applications. As an example the smart meters project has the potential to supply data from hundreds of thousands of electricity meters in real time. Apart from billing, this data could be very useful for researching energy use and reducing it. But a system will be needed to easily provide access to the data.
CSIRO ICT
The Semantic eScience Framework; toward a configurable data application format?
Peter Fox (Rensselaer Polytechnic Institute) ...
ABSTRACT:
This talk is a forward looking, technical one discussing current work including Drupal, Opendap, Virtual Observatories, Provenance, and ontology modularization, and including a summary of the keynote presented at The Australasian Ontology Workshop on December 2.
BIO:
Professor Peter Fox, now of Rensselaer Polytechnic Institute and the Tetherless World Constellation in the US (and previously of UCAR, University Corporation for Atmosperic Research) is known for his work in applications of ontologies to e-science, especially virtual observatories. He is also president of OPeNDAP which has developed standards used by NASA and NOAA to serve satellite, weather and other observed earth science data. See http://tw.rpi.edu/wiki/Peter_Fox
Peter is available for discussions and meetings from 2--4 December. Contact Kerry Taylor 6216 7038 Kerry.Taylor@csiro.au to arrange. ...
Tuesday, May 05, 2009
Future of web standards
12 to 20 May, presenting on the future of web standards, HTML5, XHTML and Semantic Technology. There are industry public and research presentations in Brisbane 12 May, Sydney 13 May, Canberra 15 May, Hobart 18 and 19 May, Melbourne 20 May. The tussle between XHTML (lead by the academics) and HTML5 (promoted by the browser writers) is an interesting one. The semantic web has similar issues.
Monday, May 05, 2008
Service Sciences and Semantics
First speaker: Dr Darrell Williamson, Deputy Director for the CSIRO ICT Centre
Topic 1: Service Science technologies and architectures in the ICT Centre
Darrell will be discussing the science, technologies and architectures that CSIRO ICT Centre has been developing in the area of Service Sciences. He will describe the role of Services in general, then follow it up with work that has been done in the area of Service Platforms, Semantic services, Web services and what the future holds in these areas.
Dr Darrell Williamson has undergraduate degrees in Science and Electrical Engineering, a Masters degree in Electrical Engineering from the University of Newcastle and a Doctoral degree in Applied Mathematics from Harvard University. Darrell was Head of Department of Engineering, and Dean of the Faculty of Engineering and Information Technology at the Australian National University, and later Director of the Telecommunications and Information Technology Research Institute at the University of Wollongong. Darrell was also Chief Executive Officer for the Cooperative Research Centre (CRC) in Advanced Computational Systems, and in February 2003, he became Chief Executive Officer for the Smart Internet Technology CRC. Darrell has served on various industries advisory committees including the Information Industries Development Board of the ACT Government and the National Advisory Committee of Item3 Pty Ltd. Darrell is currently the Deputy Director for the CSIRO ICT Centre.
Second speaker: David Ratcliffe, CSIRO e-Services Integration group
Topic 2: GRDDL - an explanation and demonstration
One of the recent recommendations that has come out of the W3C is the GRDDL (Gleaning Resource Descriptions from Dialects of Languages). David will be explaining what it is and also its application. More than that he will be walking you through a demonstration so that you can get hand's on experience with semantic technologies.
David Ratcliffe joined the e-Services Integration (eSI) group at CSIRO in 2004 as a software engineer working on data and web service composition planning for the Wedgetail (DFDMSA) and EDKMS2 projects. David was concurrently involved in upgrades to the Australian Plant Pest Database (APPD), a live information integration system, and the Boeing RFID1 project. Since 2005, David has continued work on the EDKMS projects, focusing on further developing the semantic data and web service integration capabilities of the eSI group.
Wednesday, August 29, 2007
W3C Australia Standards Symposium
World Wide Web Consortium Australia
The World Wide Web Consortium's Australian office ( W3C Aus) is run by CSIRO in Canberra (on the other side of my office wall in the ANU Computer Science and Information Technology Building).
W3C issue what they call "recommendations", but which are really standards, for HTML, XML, CSS and other key web technologies. W3C was founded by Tim Berners-Lee, inventor of the web, in 1994. As with any standards work, there is a rich mix of political, technological and commercial forces at work.
A recent area of tension touched on in the introduction was the schism in the web community between HTML and XHTML. Those working on the next version of HTML (HTML 5) have clearly stated they want to go a different direction from the work on the next XHTML (version 2).
Other tensions are with intellectual property issues with web recommendations. W3C aims to produce technology which can be freely used, without payment of royalties.
W3C wants to expand the web beyond desktop computers, to devices such as mobile phones. That probably is more a matter of commerce, than technology, but the advent of new consumer smart phones may make a differecne.
Typically the W3C process is to first have a "workshop" in an area of interest, then a working groups is formed (if justified) and publishes drafts for comment, implementations are produced to see the technology works, and after several more drafts a recommendation is released. Perhaps more importantly, W3C releases revisions and new versions of recommendations. Implementation guides and web tools are also provided to help with implementation.
As well as the more technical standards for HTML and CSS, W3C also produces guidelines, such as those for web accessibility. There are dozens of working groups working on interrelated recommendations who need to coordinate their work. W3C membership costs money and working group members contribute their time for free.
W3C Australia head, Ross Ackland, claimed the future of the web was to: semantic web, mobile web, and sensor web. He suggested we were in the middle of a ten year adoption of the mobile web, with the semantic web was further in the future and sensor web was a newly emerging technology CSIRO would like to foster.
The semantic web tries to make a web which machines can understand. Ross argued that Web 2.0 and mashups were a "grass roots" ad-hoc approach to what the semantic web was attempting. My view is that WSeb 2.0 and mashups were providing useful services, while semantic web is a failure which should be abandoned.
The W3C Mobile Web Initiative in 2005 got the attention of the mobile phone industry. But the industry has had several attempts at turning the mobile phone into a viable mobile web device. The industry's attempt with WAP was a failure costing billions of dollars. W3C's own attempt with XHTML Basic, has had limited success. About the only one to be successful was Japan's iMode, which uses a version of HTML which the W3C rejected.
The Sensor Web will provide some standards for sensor access in the future:
The Sensor Web is a type of sensor network or geographic information system (GIS) that is especially well suited for environmental monitoring and control. The term describes a specific type of sensor network: an amorphous network of spatially distributed sensor platforms (pods) that wirelessly communicate with each other. This amorphous architecture is unique since it is both synchronous and router-free, making it distinct from the more typical TCP/IP-like network schemes. The architecture allows every pod to know what is going on with every other pod throughout the Sensor Web at each measurement cycle.CSIRO have a sensor web in Brisbane which can be accessed via web services:
From: Sensor Web, Wikipedia, 21:20, 26 July 2007
Ross ended by asking what Australia could do for web standards. He pointed out that successful standards also needed market adoption. Standards take about five years to develop. The benefits are global. How does Australia contribute? An example is standards for water data standards to help with conservation in Australia and world wide.This server contains test deployments of the Open Geospatial Consortium's (OGC) Sensor Web Enablement (SWE) services. ... getCapabilities ... data from the sensors deployed by the Autonomous Systems Laboratory in Brisbane, Australia. The sensor measure temperature, soil moisture and onboard diagnostics at three locations, qcat, belmont and gatton. There are roughly 125 stations with two or three sensors each. This yields over 250 data sources of which about 150 appear to be active. Each source reports every few minutes with data coming in every few seconds. ...
From: CSIRO ICT Centre SWE Web Services, CSIRO ICT Centre, 20 April 2007
OPEN GEOSPATIAL CONSORTIUM
OGC develops "specifications" for digital maps. The aim is to be able to knit together different online mapping services to produce a coherent view for the user. OGC works with W3C groups, ISO (ISO 191xx series including ISO 19115 for Metadata) and OASIS (such as Common Alert Protocol (CAP) for emergency messages), IEEE (Sensor Model Language: SensorML).
OGC sponsors scenarios to test implementation of standards (much like the Coalition Warrior Interoperability Demonstration [CWID] for military IT). OWS 4 in December 2006 worked on sensor web enablement SWE, geo processing workflow GPN and geo-decision support. OWS 5 for 2007 is being planned.
One thing which got my attention was mention of "Social Change On-line".
At question time there was a philosophical discussion of what a standard was, their benefits, disadvantages and processes. This was entertaining but not very enlightening. Perhaps there is a need for some courses on what standards are and how they are created.
Organisation for the Advancement of Structured Information Standards
Organisation for the Advancement of Structured Information Standards (OASIS) was foundered in 1993 for SGML related standards (more recently XML standards). It has more than 60 technical committees. Individuals and organisations can join. A well known OASIS standard is ODF, based on the OpenOffice.org office document format. OASIS produces horizontal standards (general purpose technology) and vertical standards (for a particular business function). Other standards are Universal Business Language (UBL) , Customer Information Quality (CIQ) for identifying locations, organisations and people and Common Alert Protocol (CAP) for emergency messages.
Semantic Web
W3C's Semantic Web is about being able to process information. Current work is on an English-like version of the Web Ontology Language (OWL). This reminds me of the attempt with COBOL to create an English-like computer programming language which could be understood by non technical business people. The result was a verbose language which was still unintelligible to business people and cumbersome for trained computer programmers.
SPARQL is the semantic query language. POWDER the Protocol for Web Description Resources. GRDDL the Gleaning Resources Descriptions and Dialects of Languages.
This was the least useful session of the day. The Semantic Web may well turn out to be very useful one day, but so far all that appears to have been produced are a bewildering array of unintelligible standards. About the only prospect of any of this work ever being of use would be to apply the process Tim Berners-Lea used to create the web, where he took a large and complex standard (SGML) and trimmed it down to the essentials to make HTML.
GEOSCIENCE AUSTRALIA
Chris Body presented about standards in Geoscience Australia. GA seem to have suddenly become more visible, with work on geospatial standards and Tsunami warnings. The Special Minister of State, Gary Nairn, announced an Australian Spatial Consortium (ASC), on 14 August 2007, but it was not clear to me what this is.
ANZLIC (Spatial Information Council) have provided the ANZLIC Metadata Profile (December 2006) ISO TC211 framework. GeoNetwork is a metadata entry tool endorsed by Australian agencies in August 2007.
Geoscience people have a preference for formal international standards. However, GA is aiming to have any Australian contributions to be available free for public use under a Creative Commons licence.
Australian Government Information Management Office
Brian Stonebridge from AGIMO working on a standards governance framework. Brian argued that standards are boring to end users, there has to be some value to the user to get them interested. Brian's presentation was the most impressive of the day, because he was taking about how the standards could be used for the benefit of the community and he actualled used the technology he was talking about to make the presentation, via AGIMO's GovDex:
GovDex is a resource developed by government agencies to facilitate business process collaboration across policy portfolios (eg. Taxation, Human Services etc.) and administrative jurisdictions i.e. federal, state or local government levels. ...Brian mentioned that some of the work is being done online, via the system with the French government.
From: Welcome to GovDex, Australian Government Information Management Office, 2007
Brian estimated that development of standards for government use will cost about $2M a year to administer. This is not the development of new technical standards from scratch, but selecting and profiling standards for a particular application (such as selecting e-document formats for an electronic application for building a house).
AGIMO have developed a plugin for enterprise architect for government standards.
AGIMO will use underlying international and national standards, and over this methods and tools, governance and references models. The business case for this is that it will reduce the cost over time.
Unfortunately Brian then lost me in an assortment of acronyms, including:
- GIEM, Government Information Exchange Methodology (UMM v2.0 and CCTS v2.0). This extends the Canadian GSRM and is similar to the upper layers of AGA.
- AGOSP: Australian Government Online Services Portal.
Overview of the day
Ross Ackland argued that we were now "moving up the stack": the low level standards for digital communications using the Internet are set and largely working. The web provides an digital publishing overlay for this. Now more semantic content is being added to the web with standards in areas such as Geoscience and more general areas such as the Semantic Web. This is a useful way to think about the work, but the reality I see is not such a clear or systematic path.
Ross asked what should W3C and other bodies do to further standards in Australia. W3C has only a few full memebrs in Australia, due to the small size of the It industry.
I suggested that NICTA, CSIRO and other interested parties could create a one hour presentation explaining how standards development works in Australia. This could be placed on the web and offered to ACS and other IT groups to explain where standards come from and how they could get involved. This may help avoid some of the controversy and confusion surrounding issues such as the proposed adoption of Microsoft's OOXML format as an ISO standard.
One way to look at this which Ross pointed out is that the point of view about the systems are built will change: instead of building an application for an organisation and then try to interface it to other organisations, we will build the interfaces first. From the wiser perspective, I suggested that the web standards effort could be seen as building a global computer system for processing information, much as the Internet is a global system for communicating information.
Some Overall Issues on the Day
* WHERE IS ASIA?: Several speakers talked of how the standards committees were heavily influenced by US government agencies (particularly the military and security) and less so by European organisations. There appears to be little involvement by Asian organisations. There appeared to be a lack of interest in why this is so, the problems it will cause and what to do about it. Australia is culturally close to the USA and Europe and so can ride on the coat tails of the current standards process. However, at some point Asian countries and industries may decide their interests are not being served by the current standards process and decide to set up a new process for standards. Perhaps Australia can play a part in bridging the gap. This could address cultural and geopolitical issues using the web technology itself.
* USING THE STANDARDS: Many groups are producing advanced web standards. Some Internet and web tools are being used by committees. But the output of the standards committees are PDF documents or web pages. It might be useful for the web standards groups to apply some of the technology they are proposing to the standards process itself.
* USING STANDARDS: Perhaps one area in which Australia can contribute is to helping test and implement standards. This will provide useful feedback to the standards developers and also provide potential useful products.
* AUSTRALIAN DEVELOPMENT THROUGH STANDARDS: The most productive part of the day was meeting David Peterson from Boab Interactive . This Australian IT company is the latest member of W3C Australia. They are based in Townsville, North Queensland and doing web work, mostly with tropical environment research projects. Some years ago the AUstralian government funded me to see how to get regional ICT happening.
Friday, April 13, 2007
Sahana Wins Free Software Foundation Award for Social Benefit
Sahana is up to version 2 and there is work going on to add more mapping functions and mobile features. After a visit by the Sahana team to Sir Tim Berners Lee (inventor of the Web), the W3C has taken an interest. The W3C might apply its semantic web expertise to disaster management.